11 research outputs found

    HSIC Regularized LTSA

    Get PDF
    Hilbert-Schmidt Independence Criterion (HSIC) measures statistical independence between two random variables. However, instead of measuring the statistical independence between two random variables directly, HSIC first transforms two random variables into two Reproducing Kernel Hilbert Spaces (RKHS) respectively and then measures the kernelled random variables by using Hilbert-Schmidt (HS) operators between the two RKHS. Since HSIC was first proposed around 2005, HSIC has found wide applications in machine learning. In this paper, a HSIC regularized Local Tangent Space Alignment algorithm (HSIC-LTSA) is proposed. LTSA is a well-known dimensionality reduction algorithm for local homeomorphism preservation. In HSIC-LTSA, behind the objective function of LTSA, HSIC between high-dimensional and dimension-reduced data is added as a regularization term. The proposed HSIC-LTSA has two contributions. First, HSIC-LTSA implements local homeomorphism preservation and global statistical correlation during dimensionality reduction. Secondly, HSIC-LTSA proposes a new way to apply HSIC: HSIC is used as a regularization term to be added to other machine learning algorithms. The experimental results presented in this paper show that HSIC-LTSA can achieve better performance than the original LTSA

    Bid Optimization for Offsite Display Ad Campaigns on eCommerce

    Full text link
    Online retailers often use third-party demand-side-platforms (DSPs) to conduct offsite advertising and reach shoppers across the Internet on behalf of their advertisers. The process involves the retailer participating in instant auctions with real-time bidding for each ad slot of their interest. In this paper, we introduce a bid optimization system that leverages the dimensional bidding function provided by most well-known DSPs for Walmart offsite display ad campaigns. The system starts by automatically searching for the optimal segmentation of the ad requests space based on their characteristics such as geo location, time, ad format, serving website, device type, etc. Then, it assesses the quality of impressions observed from each dimension based on revenue signals driven by the campaign effect. During the campaign, the system iteratively approximates the bid landscape based on the data observed and calculates the bid adjustments for each dimension. Finally, a higher bid adjustment factor is applied to dimensions with potentially higher revenue over ad spend (ROAS), and vice versa. The initial A/B test results of the proposed optimization system has shown its effectiveness of increasing the ROAS and conversion rate while reducing the effective cost per mille for ad serving

    Change Point Detection on a Separable Model for Dynamic Networks

    Full text link
    This paper studies the change point detection problem in time series of networks, with the Separable Temporal Exponential-family Random Graph Model (STERGM). We consider a sequence of networks generated from a piecewise constant distribution that is altered at unknown change points in time. Detection of the change points can identify the discrepancies in the underlying data generating processes and facilitate downstream dynamic network analysis tasks. Moreover, the STERGM that focuses on network statistics is a flexible model to fit dynamic networks with both dyadic and temporal dependence. We propose a new estimator derived from the Alternating Direction Method of Multipliers (ADMM) and the Group Fused Lasso to simultaneously detect multiple time points, where the parameters of STERGM have changed. We also provide Bayesian information criterion for model selection to assist the detection. Our experiments show good performance of the proposed method on both simulated and real data. Lastly, we develop an R package CPDstergm to implement our method

    Real-time correction method of Muskingum model based on Kalman filter

    No full text
    In flood forecasting, general flood forecasting models or empirical forecasts reflect the average optimal value or relationship curve under the previous data. However, in the operation forecast, the forecast plan value often deviates from the actual situation. This paper takes Muskingum model as an example, and uses the Kalman filter algorithm to correct the forecast results. The algorithm structure and principles were described detailed, and the numerical simulation test was set to verify the efficiency of the Kalman filter algorithm. The correct results with corrected method were compared. The results indicated that the efficiency of the updating system using Kalman filter algorithm was improved. Conclusively, the proposed method could be widely applied in real-time flood forecast updating

    Analysis of Time and Beam Synchronization Errors for Distributed Spaceborne SAR System

    No full text
    Synchronization is a key problem in distributed Synthetic Aperture Radar (SAR) systems. In this paper, we perform a complex mathematical deduction and then analyze the influences of time synchronization on the SAR imaging and interferometric process. We discuss the relationship between time and phase synchronization, considering that different oscillators in separated transmitters and receivers lead to both time and phase synchronization errors. With respect to beam synchronization, we present the effects of the accuracies of beam pointing and satellite attitude on the antenna gain, based on the attitude-steering strategy, which involves azimuth weighting of the Doppler spectra for independent zero-Doppler beam steering. We also analyze the influences of beam synchronization on Doppler decorrelation, Signal-to-Noise Ratio (SNR), and overlapping swath error. We conduct simulations to validate the analysis results. Our findings provide guidance for system design
    corecore